Guiding CLASSY Toward More Responsive Summaries

نویسندگان

  • John M. Conroy
  • Judith D. Schlesinger
  • Peter A. Rankel
  • Dianne P. O'Leary
چکیده

We discuss changes and improvements in CLASSY for TAC 2010 along with a set of evaluation metrics. Results for both summarization and AESOP evaluation are given.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Arabic/English Multi-document Summarization with CLASSY - The Past and the Future

Automatic document summarization has become increasingly important due to the quantity of written material generated worldwide. Generating good quality summaries enables users to cope with larger amounts of information. English-document summarization is a difficult task. Yet it is not sufficient. Environmental, economic, and other global issues make it imperative for English speakers to underst...

متن کامل

CLASSY 2009: Summarization and Metrics

This year the CLASSY team participated in the update summary task and made four submissions to summarization evaluation (AESOP). Our AESOP submissions used combinations of ROUGE scores along with an update (or newness) score. We also use these new metrics, which we call Nouveau ROUGE, to help train our system and evaluate new ideas on computing update summaries. CLASSY (Clustering, Linguistics,...

متن کامل

CLASSY 2011 at TAC: Guided and Multi-lingual Summaries and Evaluation Metrics

We present CLASSY’s guided summarization as well as multi-lingual methods as submitted to TAC 2011. In addition, we describe improved metrics submitted to the AESOP task at TAC.

متن کامل

CLASSY Query-Based Multi-Document Summarization

Our summarizer is based on an HMM (Hidden Markov Model) for sentence selection within a document and a pivoted QR algorithm to generate a multi-document summary. Each year, since we began participating in DUC in 2001, we have modified the features used by the HMM and have added linguistic capabilities in order to improve the summaries we generate. Our system, called “CLASSY” (Clustering, Lingui...

متن کامل

Assessing the Effect of Inconsistent Assessors on Summarization Evaluation

We investigate the consistency of human assessors involved in summarization evaluation to understand its effect on system ranking and automatic evaluation techniques. Using Text Analysis Conference data, we measure annotator consistency based on human scoring of summaries for Responsiveness, Readability, and Pyramid scoring. We identify inconsistencies in the data and measure to what extent the...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2010